翻訳と辞書
Words near each other
・ Maxine, West Virginia
・ Maximum throughput scheduling
・ Maximum time aloft
・ Maximum time in grade
・ Maximum time interval error
・ Maximum tolerable period of disruption
・ Maximum Trance
・ Maximum transmission unit
・ Maximum Unicorn
・ Maximum usable frequency
・ Maximum Velocity (V-Max)
・ Maximum Violence
・ Maximum wage
・ Maximum Weight Convention, 1967
・ Maximum! Maximum!! Maximum!!!
Maximum-entropy Markov model
・ Maximum-Football
・ Maximum-minimums identity
・ Maximum-Residual Multicast Protocol
・ Maximum-value composite procedure
・ Maximumrocknroll
・ Maximums
・ Maximus
・ Maximus (album)
・ Maximus (BBS)
・ Maximus (Bishop of Zaragoza)
・ Maximus (comics)
・ Maximus Air Cargo
・ Maximus Dan
・ Maximus I of Constantinople


Dictionary Lists
翻訳と辞書 辞書検索 [ 開発暫定版 ]
スポンサード リンク

Maximum-entropy Markov model : ウィキペディア英語版
Maximum-entropy Markov model

In machine learning, a maximum-entropy Markov model (MEMM), or conditional Markov model (CMM), is a graphical model for sequence labeling that combines features of hidden Markov models (HMMs) and maximum entropy (MaxEnt) models. An MEMM is a discriminative model that extends a standard maximum entropy classifier by assuming that the unknown values to be learnt are connected in a Markov chain rather than being conditionally independent of each other. MEMMs find applications in natural language processing, specifically in part-of-speech tagging and information extraction.
==Model==

Suppose we have a sequence of observations O_1, \dots, O_n that we seek to tag with the labels S_1, \dots, S_nthat maximize the conditional probability P(S_1, \dots, S_n | O_1, \dots, O_n). In a MEMM, this probability is factored into Markov transition probabilities, where the probability of transitioning to a particular label depends only on the observation at that position and the previous position's label:
:P(S_1, \dots, S_n | O_1, \dots, O_n) = \prod_^nP(S_|S_,O_t).
Each of these transition probabilities come from the same general distribution P(s|s',o). For each possible label value of the previous label s', the probability of a certain label s is modeled in the same way as a maximum entropy classifier:
:P(s|s',o) = P_(s|o) = \frac\exp\left(\sum_a\lambda_af_a(o,s)\right).
Here, the f_a(o,s) are real-valued or categorical feature-functions, and Z(o,s') is a normalization term ensuring that the distribution sums to one. This form for the distribution corresponds to the maximum entropy probability distribution satisfying the constraint that the empirical expectation for the feature is equal to the expectation given the model:
: \operatorname_e\left() = \operatorname_p\left() \quad \text a .
The parameters \lambda_a can be estimated using generalized iterative scaling. Furthermore, a variant of the Baum–Welch algorithm, which is used for training HMMs, can be used to estimate parameters when training data has incomplete or missing labels.〔
The optimal state sequence S_1, \dots, S_n can be found using a very similar Viterbi algorithm to the one used for HMMs. The dynamic program uses the forward probability:
:\alpha_(s) = \sum_ \alpha_t(s') P_(s|o_).

抄文引用元・出典: フリー百科事典『 ウィキペディア(Wikipedia)
ウィキペディアで「Maximum-entropy Markov model」の詳細全文を読む



スポンサード リンク
翻訳と辞書 : 翻訳のためのインターネットリソース

Copyright(C) kotoba.ne.jp 1997-2016. All Rights Reserved.